Goto

Collaborating Authors

 transaction data


TREASURE: A Transformer-Based Foundation Model for High-Volume Transaction Understanding

Yeh, Chin-Chia Michael, Saini, Uday Singh, Dai, Xin, Fan, Xiran, Jain, Shubham, Fan, Yujie, Sun, Jiarui, Wang, Junpeng, Pan, Menghai, Dou, Yingtong, Chen, Yuzhong, Rakesh, Vineeth, Wang, Liang, Zheng, Yan, Das, Mahashweta

arXiv.org Artificial Intelligence

Payment networks form the backbone of modern commerce, generating high volumes of transaction records from daily activities. Properly modeling this data can enable applications such as abnormal behavior detection and consumer-level insights for hyper-personalized experiences, ultimately improving people's lives. In this paper, we present TREASURE, TRansformer Engine As Scalable Universal transaction Representation Encoder, a multipurpose transformer-based foundation model specifically designed for transaction data. The model simultaneously captures both consumer behavior and payment network signals (such as response codes and system flags), providing comprehensive information necessary for applications like accurate recommendation systems and abnormal behavior detection. Verified with industry-grade datasets, TREASURE features three key capabilities: 1) an input module with dedicated sub-modules for static and dynamic attributes, enabling more efficient training and inference; 2) an efficient and effective training paradigm for predicting high-cardinality categorical attributes; and 3) demonstrated effectiveness as both a standalone model that increases abnormal behavior detection performance by 111% over production systems and an embedding provider that enhances recommendation models by 104%. We present key insights from extensive ablation studies, benchmarks against production models, and case studies, highlighting valuable knowledge gained from developing TREASURE.


TransactionGPT

Dou, Yingtong, Jiang, Zhimeng, Zhang, Tianyi, Hu, Mingzhi, Xu, Zhichao, Jain, Shubham, Saini, Uday Singh, Fan, Xiran, Sun, Jiarui, Pan, Menghai, Wang, Junpeng, Dai, Xin, Wang, Liang, Yeh, Chin-Chia Michael, Fan, Yujie, Rakesh, Vineeth, Chen, Huiyuan, Bendre, Mangesh, Zhuang, Zhongfang, Li, Xiaoting, Aboagye, Prince, Lai, Vivian, Xu, Minghua, Yang, Hao, Cai, Yiwei, Das, Mahashweta, Chen, Yuzhong

arXiv.org Artificial Intelligence

TGPT is designed to understand and generate transaction trajectories while simultaneously supporting a variety of downstream prediction and classification tasks. We introduce a novel 3D-Transformer architecture specifically tailored for capturing the complex dynamics in payment transaction data. This architecture incorporates design innovations that enhance modality fusion and computational efficiency, while seamlessly enabling joint optimization with downstream objectives. Trained on billion-scale real-world transactions, TGPT significantly improves downstream classification performance against a competitive production model and exhibits advantages over baselines in generating future transactions. We conduct extensive empirical evaluations utilizing a diverse collection of company transaction datasets spanning multiple downstream tasks, thereby enabling a thorough assessment of TGPT's effectiveness and efficiency in comparison to established methodologies. Furthermore, we examine the incorporation of LLM-derived embeddings within TGPT and benchmark its performance against fine-tuned LLMs, demonstrating that TGPT achieves superior predictive accuracy as well as faster training and inference. We anticipate that the architectural innovations and practical guidelines from this work will advance foundation models for transaction-like data and catalyze future research in this emerging field.


Hybrid GCN-GRU Model for Anomaly Detection in Cryptocurrency Transactions

Na, Gyuyeon, Park, Minjung, Cha, Hyeonjeong, Kim, Soyoun, Moon, Sunyoung, Lee, Sua, Choi, Jaeyoung, Lee, Hyemin, Chai, Sangmi

arXiv.org Artificial Intelligence

Blockchain transaction networks are complex, with evolving temporal patterns and inter - node relationships. To detect illicit activi - ties, we propose a hybrid GCN - GRU model that captu res both structural and sequential features. Using real Bitcoin transaction data (2020 - 2024), our model achieved 0.9470 Accuracy and 0.9807 AUC - ROC, outperform - ing all baselines.


FinSurvival: A Suite of Large Scale Survival Modeling Tasks from Finance

Green, Aaron, Nie, Zihan, Qin, Hanzhen, Seneviratne, Oshani, Bennett, Kristin P.

arXiv.org Artificial Intelligence

Survival modeling predicts the time until an event occurs and is widely used in risk analysis; for example, it's used in medicine to predict the survival of a patient based on censored data. There is a need for large-scale, realistic, and freely available datasets for benchmarking artificial intelligence (AI) survival models. In this paper, we derive a suite of 16 survival modeling tasks from publicly available transaction data generated by lending of cryptocurrencies in Decentralized Finance (DeFi). Each task was constructed using an automated pipeline based on choices of index and outcome events. For example, the model predicts the time from when a user borrows cryptocurrency coins (index event) until their first repayment (outcome event). We formulate a survival benchmark consisting of a suite of 16 survival-time prediction tasks (FinSurvival). We also automatically create 16 corresponding classification problems for each task by thresholding the survival time using the restricted mean survival time. With over 7.5 million records, FinSurvival provides a suite of realistic financial modeling tasks that will spur future AI survival modeling research. Our evaluation indicated that these are challenging tasks that are not well addressed by existing methods. FinSurvival enables the evaluation of AI survival models applicable to traditional finance, industry, medicine, and commerce, which is currently hindered by the lack of large public datasets. Our benchmark demonstrates how AI models could assess opportunities and risks in DeFi. In the future, the FinSurvival benchmark pipeline can be used to create new benchmarks by incorporating more DeFi transactions and protocols as the use of cryptocurrency grows.


FinMaster: A Holistic Benchmark for Mastering Full-Pipeline Financial Workflows with LLMs

Jiang, Junzhe, Yang, Chang, Cui, Aixin, Jin, Sihan, Wang, Ruiyu, Li, Bo, Huang, Xiao, Sun, Dongning, Wang, Xinrun

arXiv.org Artificial Intelligence

Financial tasks are pivotal to global economic stability; however, their execution faces challenges including labor intensive processes, low error tolerance, data fragmentation, and tool limitations. Although large language models (LLMs) have succeeded in various natural language processing tasks and have shown potential in automating workflows through reasoning and contextual understanding, current benchmarks for evaluating LLMs in finance lack sufficient domain-specific data, have simplistic task design, and incomplete evaluation frameworks. To address these gaps, this article presents FinMaster, a comprehensive financial benchmark designed to systematically assess the capabilities of LLM in financial literacy, accounting, auditing, and consulting. Specifically, FinMaster comprises three main modules: i) FinSim, which builds simulators that generate synthetic, privacy-compliant financial data for companies to replicate market dynamics; ii) FinSuite, which provides tasks in core financial domains, spanning 183 tasks of various types and difficulty levels; and iii) FinEval, which develops a unified interface for evaluation. Extensive experiments over state-of-the-art LLMs reveal critical capability gaps in financial reasoning, with accuracy dropping from over 90% on basic tasks to merely 40% on complex scenarios requiring multi-step reasoning. This degradation exhibits the propagation of computational errors, where single-metric calculations initially demonstrating 58% accuracy decreased to 37% in multimetric scenarios. To the best of our knowledge, FinMaster is the first benchmark that covers full-pipeline financial workflows with challenging tasks. We hope that FinMaster can bridge the gap between research and industry practitioners, driving the adoption of LLMs in real-world financial practices to enhance efficiency and accuracy.


Synthetic Data for Robust AI Model Development in Regulated Enterprises

Godbole, Aditi

arXiv.org Artificial Intelligence

In today's business landscape, organizations need to find the right balance between using their customers' data ethically to power AI solutions and being compliant regarding data privacy and data usage regulations. In this paper, we discuss synthetic data as a possible solution to this dilemma. Synthetic data is simulated data that mimics the real data. We explore how organizations in heavily regulated industries, such as financial institutions or healthcare organizations, can leverage synthetic data to build robust AI solutions while staying compliant. We demonstrate that synthetic data offers two significant advantages by allowing AI models to learn from more diverse data and by helping organizations stay compliant against data privacy laws with the use of synthetic data instead of customer information. We discuss case studies to show how synthetic data can be effectively used in the finance and healthcare sector while discussing the challenges of using synthetic data and some ethical questions it raises. Our research finds that synthetic data could be a game-changer for AI in regulated industries. The potential can be realized when industry, academia, and regulators collaborate to build solutions. We aim to initiate discussions on the use of synthetic data to build ethical, responsible, and effective AI systems in regulated enterprise industries.


Instruction-Based Fine-tuning of Open-Source LLMs for Predicting Customer Purchase Behaviors

Ergul, Halil Ibrahim, Balcisoy, Selim, Bozkaya, Burcin

arXiv.org Artificial Intelligence

In this study, the performance of various predictive models, including probabilistic baseline, CNN, LSTM, and finetuned LLMs, in forecasting merchant categories from financial transaction data have been evaluated. Utilizing datasets from Bank A for training and Bank B for testing, the superior predictive capabilities of the fine-tuned Mistral Instruct model, which was trained using customer data converted into natural language format have been demonstrated. The methodology of this study involves instruction fine-tuning Mistral via LoRA (LowRank Adaptation of Large Language Models) to adapt its vast pre-trained knowledge to the specific domain of financial transactions. The Mistral model significantly outperforms traditional sequential models, achieving higher F1 scores in the three key merchant categories of bank transaction data (grocery, clothing, and gas stations) that is crucial for targeted marketing campaigns. This performance is attributed to the model's enhanced semantic understanding and adaptability which enables it to better manage minority classes and predict transaction categories with greater accuracy. These findings highlight the potential of LLMs in predicting human behavior.


Dynamic Feature Fusion: Combining Global Graph Structures and Local Semantics for Blockchain Fraud Detection

Sheng, Zhang, Song, Liangliang, Wang, Yanbin

arXiv.org Artificial Intelligence

--The advent of blockchain technology has facilitated the widespread adoption of smart contracts in the financial sector . However, current fraud detection methodologies exhibit limitations in capturing both global structural patterns within transaction networks and local semantic relationships embedded in transaction data. Most existing models focus on either structural information or semantic features individually, leading to suboptimal performance in detecting complex fraud patterns.In this paper, we propose a dynamic feature fusion model that combines graph-based representation learning and semantic feature extraction for blockchain fraud detection. Specifically, we construct global graph representations to model account relationships and extract local contextual features from transaction data. A dynamic multimodal fusion mechanism is introduced to adaptively integrate these features, enabling the model to capture both structural and semantic fraud patterns effectively. We further develop a comprehensive data processing pipeline, including graph construction, temporal feature enhancement, and text preprocessing. Experimental results on large-scale real-world blockchain datasets demonstrate that our method outperforms existing benchmarks across accuracy, F1 score, and recall metrics. This work highlights the importance of integrating structural relationships and semantic similarities for robust fraud detection and offers a scalable solution for securing blockchain systems. LOCKCHAIN technology has developed rapidly in recent years and has triggered far-reaching changes in several fields, especially in the financial industry [1]. However, as the popularity of blockchain applications grows, so does the significant increase in fraudulent behaviors it has brought about, with serious implications for society [2]. Blockchain technology, due to its decentralization and transparency, has become a tool for unscrupulous individuals to exploit, although it provides greater security and efficiency in financial transactions [3]. For example, the application of blockchain technology in the supply chain is seen as an effective means to enhance transparency and traceability, but it also faces a crisis of social trust due to fraudulent behavior [4].


Generative AI for Banks: Benchmarks and Algorithms for Synthetic Financial Transaction Data

Karst, Fabian Sven, Chong, Sook-Yee, Antenor, Abigail A., Lin, Enyu, Li, Mahei Manhai, Leimeister, Jan Marco

arXiv.org Artificial Intelligence

The banking sector, as a data-driven industry, relies on the availability of high-quality data to create value and protect its customers. The synergy between recent deep learning (DL) advancements, and the sector's data needs presents a growth potential of USD$4.6 trillion by 2035 (Accenture, 2017). However, deploying DL models is challenging due to the need for large, high-quality training data (Ryll et al., 2020), a difficulty made worse by the intricacy of financial transaction data (with complex data patterns and time-related characteristics), and strict regulations that limit data sharing (EU Regulation 2016/679, PCI DSS v4.0). One possible solution is to use synthetic data which is artificially generated rather than drawn from real-world events to increase samples in the minority class (Jordon et al., 2022), and allow safe data sharing between financial institutions while protecting privacy (Karst et al., 2024). This approach is essential for improving models used in assessing risks and detecting fraud.


Blockchain Data Analysis in the Era of Large-Language Models

Toyoda, Kentaroh, Wang, Xiao, Li, Mingzhe, Gao, Bo, Wang, Yuan, Wei, Qingsong

arXiv.org Artificial Intelligence

Blockchain data analysis is essential for deriving insights, tracking transactions, identifying patterns, and ensuring the integrity and security of decentralized networks. It plays a key role in various areas, such as fraud detection, regulatory compliance, smart contract auditing, and decentralized finance (DeFi) risk management. However, existing blockchain data analysis tools face challenges, including data scarcity, the lack of generalizability, and the lack of reasoning capability. We believe large language models (LLMs) can mitigate these challenges; however, we have not seen papers discussing LLM integration in blockchain data analysis in a comprehensive and systematic way. This paper systematically explores potential techniques and design patterns in LLM-integrated blockchain data analysis. We also outline prospective research opportunities and challenges, emphasizing the need for further exploration in this promising field. This paper aims to benefit a diverse audience spanning academia, industry, and policy-making, offering valuable insights into the integration of LLMs in blockchain data analysis.